Goto

Collaborating Authors

 cosmological model


Inferring Cosmological Parameters with Evidential Physics-Informed Neural Networks

Tan, Hai Siong

arXiv.org Artificial Intelligence

We examine the use of a novel variant of Physics-Informed Neural Networks to predict cosmological parameters from recent supernovae and baryon acoustic oscillations (BAO) datasets. Our machine learning framework generates uncertainty estimates for target variables and the inferred unknown parameters of the underlying PDE descriptions. Built upon a hybrid of the principles of Evidential Deep Learning, Physics-Informed Neural Networks, Bayesian Neural Networks and Gaussian Processes, our model enables learning of the posterior distribution of the unknown PDE parameters through standard gradient-descent based training. We apply our model to an up-to-date BAO dataset (Bousis et al. 2024) calibrated with the CMB-inferred sound horizon, and the Pantheon$+$ Sne Ia distances (Scolnic et al. 2018), examining the relative effectiveness and mutual consistency among the standard $Λ$CDM, $w$CDM and $Λ_s$CDM models. Unlike previous results arising from the standard approach of minimizing an appropriate $χ^2$ function, the posterior distributions for parameters in various models trained purely on Pantheon$+$ data were found to be largely contained within the $2σ$ contours of their counterparts trained on BAO data. Their posterior medians for $h_0$ were within about $2σ$ of one another, indicating that our machine learning-guided approach provides a different measure of the Hubble tension.


Deep Learning Based Recalibration of SDSS and DESI BAO Alleviates Hubble and Clustering Tensions

Shah, Rahul, Mukherjee, Purba, Saha, Soumadeep, Garain, Utpal, Pal, Supratik

arXiv.org Artificial Intelligence

Conventional calibration of Baryon Acoustic Oscillations (BAO) data relies on estimation of the sound horizon at drag epoch $r_d$ from early universe observations by assuming a cosmological model. We present a recalibration of two independent BAO datasets, SDSS and DESI, by employing deep learning techniques for model-independent estimation of $r_d$, and explore the impacts on $\Lambda$CDM cosmological parameters. Significant reductions in both Hubble ($H_0$) and clustering ($S_8$) tensions are observed for both the recalibrated datasets. Moderate shifts in some other parameters hint towards further exploration of such data-driven approaches.


Model-agnostic basis functions for the 2-point correlation function of dark matter in linear theory

Paranjape, Aseem, Sheth, Ravi K.

arXiv.org Artificial Intelligence

We consider approximating the linearly evolved 2-point correlation function (2pcf) of dark matter $\xi_{\rm lin}(r;\boldsymbol{\theta})$ in a cosmological model with parameters $\boldsymbol{\theta}$ as the linear combination $\xi_{\rm lin}(r;\boldsymbol{\theta})\approx\sum_i\,b_i(r)\,w_i(\boldsymbol{\theta})$, where the functions $\mathcal{B}=\{b_i(r)\}$ form a $\textit{model-agnostic basis}$ for the linear 2pcf. This decomposition is important for model-agnostic analyses of the baryon acoustic oscillation (BAO) feature in the nonlinear 2pcf of galaxies that fix $\mathcal{B}$ and leave the coefficients $\{w_i\}$ free. To date, such analyses have made simple but sub-optimal choices for $\mathcal{B}$, such as monomials. We develop a machine learning framework for systematically discovering a $\textit{minimal}$ basis $\mathcal{B}$ that describes $\xi_{\rm lin}(r)$ near the BAO feature in a wide class of cosmological models. We use a custom architecture, denoted $\texttt{BiSequential}$, for a neural network (NN) that explicitly realizes the separation between $r$ and $\boldsymbol{\theta}$ above. The optimal NN trained on data in which only $\{\Omega_{\rm m},h\}$ are varied in a $\textit{flat}$ $\Lambda$CDM model produces a basis $\mathcal{B}$ comprising $9$ functions capable of describing $\xi_{\rm lin}(r)$ to $\sim0.6\%$ accuracy in $\textit{curved}$ $w$CDM models varying 7 parameters within $\sim5\%$ of their fiducial, flat $\Lambda$CDM values. Scales such as the peak, linear point and zero-crossing of $\xi_{\rm lin}(r)$ are also recovered with very high accuracy. We compare our approach to other compression schemes in the literature, and speculate that $\mathcal{B}$ may also encompass $\xi_{\rm lin}(r)$ in modified gravity models near our fiducial $\Lambda$CDM model. Using our basis functions in model-agnostic BAO analyses can potentially lead to significant statistical gains.


LADDER: Revisiting the Cosmic Distance Ladder with Deep Learning Approaches and Exploring its Applications

Shah, Rahul, Saha, Soumadeep, Mukherjee, Purba, Garain, Utpal, Pal, Supratik

arXiv.org Artificial Intelligence

ABSTRACT We investigate the prospect of reconstructing the "cosmic distance ladder" of the Universe using a novel deep learning framework called LADDER - Learning Algorithm for Deep Distance Estimation and Reconstruction. LADDER is trained on the apparent magnitude data from the Pantheon Type Ia supernovae compilation, incorporating the full covariance information among data points, to produce predictions along with corresponding errors. After employing several validation tests with a number of deep learning models, we pick LADDER as the best performing one. We then demonstrate applications of our method in the cosmological context, that include serving as a model-independent tool for consistency checks for other datasets like baryon acoustic oscillations, calibration of high-redshift datasets such as gamma ray bursts, use as a model-independent mock catalog generator for future probes, etc. INTRODUCTION Knowledge of accurate distances to astronomical entities at various redshifts is essential for deducing the expansion history of the Universe. Observationally, however, this task is not simple since there does not exist one single standardizable measure of distances at all scales of cosmological interest. Hence one has to resort to a progressive method of calibrating distances, called the "cosmic distance ladder" method, using overlapping regions of potentially different standardizable objects as "rungs of the ladder". The conventional distance ladder method (Riess & Breuval 2023) starts with direct measures of geometric distance measures and progresses to calibrating Cepheid variables (Freedman & Madore 2023) or Tip of the Red Giant Branch (TRGB) stars (Freedman et al. 2020), and finally Type Ia supernovae (SNIa).

  Country: Asia > India > West Bengal > Kolkata (0.05)
  Genre: Research Report (1.00)

Amortized Simulation-Based Frequentist Inference for Tractable and Intractable Likelihoods

Kadhim, Ali Al, Prosper, Harrison B., Prosper, Olivia F.

arXiv.org Machine Learning

High-fidelity simulators that connect theoretical models with observations are indispensable tools in many sciences. When coupled with machine learning, a simulator makes it possible to infer the parameters of a theoretical model directly from real and simulated observations without explicit use of the likelihood function. This is of particular interest when the latter is intractable. In this work, we introduce a simple extension of the recently proposed likelihood-free frequentist inference (LF2I) approach that has some computational advantages. Like LF2I, this extension yields provably valid confidence sets in parameter inference problems in which a high-fidelity simulator is available. The utility of our algorithm is illustrated by applying it to three pedagogically interesting examples: the first is from cosmology, the second from high-energy physics and astronomy, both with tractable likelihoods, while the third, with an intractable likelihood, is from epidemiology.


Neural network reconstruction of cosmology using the Pantheon compilation

Dialektopoulos, Konstantinos F., Mukherjee, Purba, Said, Jackson Levi, Mifsud, Jurgen

arXiv.org Artificial Intelligence

In this work, we reconstruct the Hubble diagram using various data sets, including correlated ones, in Artificial Neural Networks (ANN). Using ReFANN, that was built for data sets with independent uncertainties, we expand it to include non-Guassian data points, as well as data sets with covariance matrices among others. Furthermore, we compare our results with the existing ones derived from Gaussian processes and we also perform null tests in order to test the validity of the concordance model of cosmology.


Reconstructing the Hubble parameter with future Gravitational Wave missions using Machine Learning

Mukherjee, Purba, Shah, Rahul, Bhaumik, Arko, Pal, Supratik

arXiv.org Artificial Intelligence

We study the prospects of Gaussian processes (GP), a machine learning (ML) algorithm, as a tool to reconstruct the Hubble parameter $H(z)$ with two upcoming gravitational wave missions, namely the evolved Laser Interferometer Space Antenna (eLISA) and the Einstein Telescope (ET). Assuming various background cosmological models, the Hubble parameter has been reconstructed in a non-parametric manner with the help of GP using realistically generated catalogs for each mission. The effects of early-time and late-time priors on the reconstruction of $H(z)$, and hence on the Hubble constant ($H_0$), have also been focused on separately. Our analysis reveals that GP is quite robust in reconstructing the expansion history of the Universe within the observational window of the specific missions under consideration. We further confirm that both eLISA and ET would be able to provide constraints on $H(z)$ and $H_0$ which would be competitive to those inferred from current datasets. In particular, we observe that an eLISA run of $\sim10$-year duration with $\sim80$ detected bright siren events would be able to constrain $H_0$ as good as a $\sim3$-year ET run assuming $\sim 1000$ bright siren event detections. Further improvement in precision is expected for longer eLISA mission durations such as a $\sim15$-year time-frame having $\sim120$ events. Lastly, we discuss the possible role of these future gravitational wave missions in addressing the Hubble tension, for each model, on a case-by-case basis.


An AI-assisted analysis of three-dimensional galaxy distribution in our universe

#artificialintelligence

By applying a machine-learning technique, a neural network method, to gigantic amounts of simulation data about the formation of cosmic structures in the universe, a team of researchers has developed a very fast and highly efficient software program that can make theoretical predictions about structure formation. By comparing model predictions to actual observational datasets, the team succeeded in accurately measuring cosmological parameters, reports a study in Physical Review D. When the biggest galaxy survey to date in the world, the Sloan Digital Sky Survey (SDSS), created a three-dimensional map of the universe via the observed distribution of galaxies, it became clear that galaxies had certain characteristics. Some would clump together, or spread out in filaments, and in some places there were voids where no galaxies existed at all. All these show galaxies did not evolve in a uniform way, they formed as a result of their local environment. In general, researchers agree this non-uniform distribution of galaxies is because of the effects of gravity caused by the distribution of "invisible" dark matter, the mysterious matter that no one has yet directly observed. By studying the data in the three-dimensional map of galaxies in detail, researchers could uncover the fundamental quantities such as the amount of dark matter in the universe.


Decoding Cosmological Information in Weak-Lensing Mass Maps with Generative Adversarial Networks

Shirasaki, Masato, Yoshida, Naoki, Ikeda, Shiro, Oogi, Taira, Nishimichi, Takahiro

arXiv.org Machine Learning

Galaxy imaging surveys enable us to map the cosmic matter density field through weak gravitational lensing analysis. The density reconstruction is compromised by a variety of noise originating from observational conditions, galaxy number density fluctuations, and intrinsic galaxy properties. We propose a deep-learning approach based on generative adversarial networks (GANs) to reduce the noise in the weak lensing map under realistic conditions. We perform image-to-image translation using conditional GANs in order to produce noiseless lensing maps using the first-year data of the Subaru Hyper Suprime-Cam (HSC) survey. We train the conditional GANs by using 30000 sets of mock HSC catalogs that directly incorporate observational effects. We show that an ensemble learning method with GANs can reproduce the one-point probability distribution function (PDF) of the lensing convergence map within a $0.5-1\sigma$ level. We use the reconstructed PDFs to estimate a cosmological parameter $S_{8} = \sigma_{8}\sqrt{\Omega_{\rm m0}/0.3}$, where $\Omega_{\rm m0}$ and $\sigma_{8}$ represent the mean and the scatter in the cosmic matter density. The reconstructed PDFs place tighter constraint, with the statistical uncertainty in $S_8$ reduced by a factor of $2$ compared to the noisy PDF. This is equivalent to increasing the survey area by $4$ without denoising by GANs. Finally, we apply our denoising method to the first-year HSC data, to place $2\sigma$-level cosmological constraints of $S_{8} < 0.777 \, ({\rm stat}) + 0.105 \, ({\rm sys})$ and $S_{8} < 0.633 \, ({\rm stat}) + 0.114 \, ({\rm sys})$ for the noisy and denoised data, respectively.


Non-Gaussian information from weak lensing data via deep learning

Gupta, Arushi, Matilla, José Manuel Zorrilla, Hsu, Daniel, Haiman, Zoltán

arXiv.org Machine Learning

Weak lensing maps contain information beyond two-point statistics on small scales. Much recent work has tried to extract this information through a range of different observables or via nonlinear transformations of the lensing field. Here we train and apply a 2D convolutional neural network to simulated noiseless lensing maps covering 96 different cosmological models over a range of {$\Omega_m,\sigma_8$}. Using the area of the confidence contour in the {$\Omega_m,\sigma_8$} plane as a figure-of-merit, derived from simulated convergence maps smoothed on a scale of 1.0 arcmin, we show that the neural network yields $\approx 5 \times$ tighter constraints than the power spectrum, and $\approx 4 \times$ tighter than the lensing peaks. Such gains illustrate the extent to which weak lensing data encode cosmological information not accessible to the power spectrum or even non-Gaussian statistics such as lensing peaks.